Bayesian K-Means as a “Maximization-Expectation” Algorithm
ثبت نشده
چکیده
We introduce a new class of “maximization expectation” (ME) algorithms where we maximize over hidden variables but marginalize over random parameters. This reverses the roles of expectation and maximization in the classical EM algorithm. In the context of clustering, we argue that these hard assignments open the door to very fast implementations based on data-structures such as kdtrees and conga-lines. The marginalization over parameters ensures that we retain the ability to infer model structure (i.e. number of clusters). As an important example we discuss a top-down “Bayesian k-means” algorithm and a bottom-up agglomerative clustering algorithm. In experiments we compare these algorithms against a number of alternative algorithms that have recently appeared in the literature.
منابع مشابه
Bayesian K-Means as a "Maximization-Expectation" Algorithm
We introduce a new class of "maximization-expectation" (ME) algorithms where we maximize over hidden variables but marginalize over random parameters. This reverses the roles of expectation and maximization in the classical expectation-maximization algorithm. In the context of clustering, we argue that these hard assignments open the door to very fast implementations based on data structures su...
متن کاملUnsupervised training of Bayesian networks for data clustering
This paper presents a new approach to the unsupervised training of Bayesian network classifiers. Three models have been analysed: the Chow and Liu (CL) multinets; the treeaugmented naive Bayes; and a new model called the simple Bayesian network classifier, which is more robust in its structure learning. To perform the unsupervised training of these models, the classification maximum likelihood ...
متن کاملClustering Methods for Credit Card using
K-means clustering algorithm is a method of cluster analysis which aims to partition n observations into clusters in which each observation belongs to the cluster with the nearest mean. It is one of the simplest unconfirmed learning algorithms that solve the well known clustering problem. It is similar to the hope maximization algorithm for mixtures of Gaussians in that they both attempt to fin...
متن کاملFinding hypergraph communities: a Bayesian approach and variational solution
Data clustering, including problems such as finding network communities, can be put into a systematic framework by means of a Bayesian approach. Here we address the Bayesian formulation of the problem of finding hypergraph communities. We start by introducing a hypergraph generative model with a built-in group structure. Using a variational calculation we derive a variational Bayes algorithm, a...
متن کاملNoise Benefits in Expectation-Maximization Algorithms
This dissertation shows that careful injection of noise into sample data can substantially speed up Expectation-Maximization algorithms. Expectation-Maximization algorithms are a class of iterative algorithms for extracting maximum likelihood estimates from corrupted or incomplete data. The convergence speed-up is an example of a noise benefit or"stochastic resonance"in statistical signal proce...
متن کامل